Cautious weighted random forests
نویسندگان
چکیده
Random forest is an efficient and accurate classification model, which makes decisions by aggregating a set of trees, either voting or averaging class posterior probability estimates. However, tree outputs may be unreliable in presence scarce data. The imprecise Dirichlet model (IDM) provides workaround, replacing point estimates with interval-valued ones. This paper investigates new aggregation method based on the theory belief functions to combine such intervals, resulting cautious random classifier. In particular, we propose strategy for computing weights minimization convex cost function, takes both determinacy accuracy into account it possible adjust level cautiousness model. proposed evaluated 25 UCI datasets demonstrated more adaptive noise training data achieve better compromise between informativeness cautiousness. • A classifier forests proposed. Tree are automatically learned from using function. tuned single parameter. Extensive experiments demonstrate interests method.
منابع مشابه
News Articles Classification Using Random Forests and Weighted Multimodal Features
This research investigates the problem of news articles classification. The classification is performed using N-gram textual features extracted from text and visual features generated from one representative image. The application domain is news articles written in English that belong to four categories: Business-Finance, Lifestyle-Leisure, Science-Technology and Sports downloaded from three we...
متن کاملHybrid weighted random forests for classifying very high-dimensional data
Random forests are a popular classification method based on an ensemble of a single type of decision trees from subspaces of data. In the literature, there are many different types of decision tree algorithms, including C4.5, CART, and CHAID. Each type of decision tree algorithm may capture different information and structure. This paper proposes a hybrid weighted random forest algorithm, simul...
متن کامل1 Random Forests - - Random Features
Random forests are a combination of tree predictors such that each tree depends on the values of a random vector sampled independently and with the same distribution for all trees in the forest. The generalization error for forests converges a.s. to a limit as the number of trees in the forest becomes large. The error of a forest of tree classifiers depends on the strength of the individual tre...
متن کاملMondrian Forests: Efficient Online Random Forests
Ensembles of randomized decision trees, usually referred to as random forests, are widely used for classification and regression tasks in machine learning and statistics. Random forests achieve competitive predictive performance and are computationally efficient to train and test, making them excellent candidates for real-world prediction tasks. The most popular random forest variants (such as ...
متن کاملRandom Composite Forests
We introduce a broad family of decision trees, Composite Trees, whose leaf classifiers are selected out of a hypothesis set composed of p subfamilies with different complexities. We prove new data-dependent learning guarantees for this family in the multi-class setting. These learning bounds provide a quantitative guidance for the choice of the hypotheses at each leaf. Remarkably, they depend o...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: Expert Systems With Applications
سال: 2023
ISSN: ['1873-6793', '0957-4174']
DOI: https://doi.org/10.1016/j.eswa.2022.118883